SelfCCL: Curriculum Contrastive Learning by Transferring Self-Taught Knowledge for Fine-Tuning BERT

نویسندگان

چکیده

BERT, the most popular deep learning language model, has yielded breakthrough results in various NLP tasks. However, semantic representation space learned by BERT property of anisotropy. Therefore, needs to be fine-tuned for certain downstream tasks such as Semantic Textual Similarity (STS). To overcome this problem and improve sentence space, some contrastive methods have been proposed fine-tuning BERT. existing models do not consider importance input triplets terms easy hard negatives during training. In paper, we propose SelfCCL: Curriculum Contrastive Learning model Transferring Self-taught Knowledge Fine-Tuning which mimics two ways that humans learn about world around them, namely curriculum learning. The former learns contrasting similar dissimilar samples. latter is inspired way from simplest concepts complex concepts. Our also performs training transferring self-taught knowledge. That is, figures out are or difficult based on previously knowledge, then those order using a objective. We apply our Sentence BERT(SBERT) frameworks. evaluation SelfCCL standard STS SentEval transfer show together with increases average performance extent.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Risk Bounds for Transferring Representations With and Without Fine-Tuning

A popular machine learning strategy is the transfer of a representation (i.e. a feature extraction function) learned on a source task to a target task. Examples include the re-use of neural network weights or word embeddings. We develop sufficient conditions for the success of this approach. If the representation learned from the source task is fixed, we identify conditions on how the tasks rel...

متن کامل

An Interactive Knowledge-based Learning Environment for Transferring Health Knowledge

This paper describes a narrative based interactive intelligent learning environment which aims to elucidate reasoning using interactive scenarios that may be used in training novices ICU nurses in the application of their knowledge and decision making. Its design is based on an approach to generating narrative from knowledge that has been modelled in specific decision/reasoning domains. The app...

متن کامل

Robust and Discriminative Self-Taught Learning

The lack of training data is a common challenge in many machine learning problems, which is often tackled by semi-supervised learning methods or transfer learning methods. The former requires unlabeled images from the same distribution as the labeled ones and the latter leverages labeled images from related homogenous tasks. However, these restrictions often cannot be satisfied. To address this...

متن کامل

Self-Paced Curriculum Learning

Curriculum learning (CL) or self-paced learning (SPL) represents a recently proposed learning regime inspired by the learning process of humans and animals that gradually proceeds from easy to more complex samples in training. The two methods share a similar conceptual learning paradigm, but differ in specific learning schemes. In CL, the curriculum is predetermined by prior knowledge, and rema...

متن کامل

Deep Self-Taught Learning for Handwritten Character Recognition

Recent theoretical and empirical work in statistical machine learning has demonstrated the importance of learning algorithms for deep architectures, i.e., function classes obtained by composing multiple non-linear transformations. Self-taught learning (exploiting unlabeled examples or examples from other distributions) has already been applied to deep learners, but mostly to show the advantage ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Applied sciences

سال: 2023

ISSN: ['2076-3417']

DOI: https://doi.org/10.3390/app13031913